home
***
CD-ROM
|
disk
|
FTP
|
other
***
search
/
Amiga Format CD 41
/
Amiga Format CD41 (1999-06)(Future Publishing)(GB)[!][issue 1999-07].iso
/
-seriously_amiga-
/
comms
/
www
/
getallhtml
/
getallhtml.readme
< prev
next >
Wrap
Text File
|
1999-04-19
|
2KB
|
57 lines
Short: Auto-download/check entire web sites! (v0.64ß)
Author: Chris.S.Handley@BTInternet.com
Uploader: Chris.S.Handley@BTInternet.com
Type: comm/tcp
Requires: HTTPResume v1.3+, Rexxsupport.library, ARexx
Version: v0.64ß
Introduction
------------
Have you ever visited a cool web site & wanted to keep a copy of some/all of it,
but it would takes ages to find & download all the respective pages/files?
This is the answer!
You supply this ARexx script with the start page URL, and a destination
directory (which should be empty), and maybe a few other options - and off it
goes! Note that it needs HTTPResume v1.3+ to work (get from Aminet).
Latest News
-----------
A fix I did in v0.61 was actually wrong - undone so that all downloading should
work properly now. Also improved BROKENLINKS & other minor things. I actually
had time to test this release, so it should work pretty well! :-)
Many people have been having problems with GetAllHTML after editing it - seems
this is due to spurious ASCII-27 characters mucking-up some editors :-( .
Anyway, I wrote a program to detect & remove all non-visible characters
(available if wanted), and it seems that GetAllHTML is the only recent text file
I wrote which had the problem... Any ideas WHY they appeared? I use CygnusEd
v3.5.
I've programmed the BROKENLINKS switch to allow web page makers to automagically
search their site for broken links - written just for Alexander Niven-Jenkins
(emailing me can be worth it;-)
Changed the NOPAUSE switch to PAUSE, so that it defaults to NOT pausing.
Very minor enhancments & fixed an arguments interpreting bug.
I will still fix major bugs until I have an AmigaE version that can be tested.
History
-------
v0.64ß (04-04-99) - Put back the 'extra' END that I removed in v0.61 . Now
BROKENLINKS will always only try to download external links
once. Removed NOENV argument of HTTPResume so proxy
settings may work. Minor changes.
v0.63ß (04-04-99) - Removed spurious non-visible ASCII (27) characters that
caused some text editors to go loopy.
v0.62ß (03-04-99) - Add the BROKENLINKS switch. Replaced NOPAUSE by PAUSE
switch. Now always warns if a file could not be downloaded
(not just pages). If you used all the arguments then it
would miss the last one.
v0.61ß (28-03-99) - Possible fix for RESUME problem done, plus stupidly left an
extra END where it broke GetAllHTML.
<snip>